# Local Inference

INTELLECT 2 GGUF
INTELLECT-2-GGUF is the GGUF format quantized version of PrimeIntellect/INTELLECT-2, suitable for text generation tasks.
Large Language Model
I
MaziyarPanahi
88
1
Qwen3 30B A3B MLX 4bit
Apache-2.0
Qwen3-30B-A3B-MLX-8bit is an 8-bit quantized version converted from Qwen/Qwen3-30B-A3B, optimized for the MLX framework and suitable for text generation tasks.
Large Language Model
Q
lmstudio-community
4,199
19
Gemma 3 27b It GGUF
GGUF quantized version of the Gemma 3 27B instruction-tuned model, suitable for text generation tasks.
Large Language Model
G
MaziyarPanahi
475.91k
4
Gemma 3 12b It GGUF
GGUF quantized version of Gemma 3 12B, suitable for text generation tasks.
Large Language Model
G
MaziyarPanahi
641.41k
4
Gemma 3 4b It GGUF
GGUF quantized version of the Gemma 3B model, suitable for local text generation tasks
Large Language Model
G
MaziyarPanahi
358.91k
6
Llama 3.2 3B Instruct Abliterated GGUF
GGUF-format quantized version of Llama-3.2-3B-Instruct-abliterated, supporting multiple bit quantization options, suitable for text generation tasks.
Large Language Model
L
MaziyarPanahi
181
2
Phi 3.5 Mini Instruct GGUF
GGUF format model file for Phi-3.5-mini-instruct, suitable for text generation tasks.
Large Language Model
P
MaziyarPanahi
335.88k
13
Mistral Nemo Instruct 2407 GGUF
Mistral-Nemo-Instruct-2407-GGUF is the GGUF format quantized version of mistralai/Mistral-Nemo-Instruct-2407, supporting multiple quantization bits (2-bit to 8-bit), suitable for text generation tasks.
Large Language Model
M
MaziyarPanahi
441.17k
44
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase